This paper introduces UniTouch, a unified tactile model that connects touch sensors to vision, language, and sound. It aligns tactile embeddings with pretrained image embeddings already linked to other modalities. Sensor-specific tokens allow learning from multiple heterogeneous tactile sensors together. Applications include zero-shot robot grasping prediction and touch...